Accelerate the Convergence Speed of Perceptron Learning Algorithm with Weight

نویسندگان

چکیده

The main problems of the traditional perceptron learning algorithm (PLA) is that there are too many iterations and it difficult to generate a model quickly, more needed when boundary between two classes closed. In this paper, we improve PLA by introducing current weight into updating formulation, which can significantly accelerate iteration. experiments on different public datasets show our proposed method greatly speed PLA.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Convergence Properties and Stationary Points of a Perceptron Learning Algorithm

The Perceptron i s an adaptive linear combiner that has its output quantized to one o f two possible discrete values, and i t is the basic component of multilayer, feedforward neural networks. The leastmean-square (LMS) adaptive algorithm adjusts the internal weights to train the network to perform some desired function, such as pattern recognition. In this paper, we present an analysis o f the...

متن کامل

Conditional convergence of photorefractive perceptron learning.

We consider the convergence characteristics of a perceptron learning algorithm, taking into account the decay of photorefractive holograms during the process of interconnection weight changes. As a result of the hologram erasure, the convergence of the learning process is dependent on the exposure time during the weight changes. A mathematical proof of the conditional convergence, perceptrons, ...

متن کامل

The Role of Weight Shrinking in Large Margin Perceptron Learning

We introduce into the classical perceptron algorithm with margin a mechanism that shrinks the current weight vector as a first step of the update. If the shrinking factor is constant the resulting algorithm may be regarded as a margin-error-driven version of NORMA with constant learning rate. In this case we show that the allowed strength of shrinking depends on the value of the maximum margin....

متن کامل

the effects of integrating cooperative learning into vocabulary learning of elementary school students

the purpose of the research is to examine if integrating cooperative learning into vocabulary learning helps to increase word recognition of students in an elementary school in iran. it tries to investigate whether cooperative learning approach enables students to improve their language learning. this research used stad (students team achievement division) as a cooperative model in this study. ...

15 صفحه اول

Multilinear perceptron convergence theorem.

An adaptive perceptron with multilinear couplings is introduced. While an adap-tive perceptron exhibits severe shortcomings if it is applied to complex tasks, this is not so for the adaptive multilinear perceptron.

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Frontiers in artificial intelligence and applications

سال: 2022

ISSN: ['1879-8314', '0922-6389']

DOI: https://doi.org/10.3233/faia220385